Posterior Contraction in Sparse Bayesian Factor Models for Massive Covariance
نویسندگان
چکیده
Sparse Bayesian factor models are routinely implemented for parsimonious dependence modeling and dimensionality reduction in highdimensional applications. We provide theoretical understanding of such Bayesian procedures in terms of posterior convergence rates in inferring high-dimensional covariance matrices where the dimension can be potentially larger than the sample size. Under relevant sparsity assumptions on the true covariance matrix, we show that commonlyused point mass mixture priors on the factor loadings lead to consistent estimation in the operator norm even when p n. One of our major contributions is to develop a new class of continuous shrinkage priors and provide insights into their concentration around sparse vectors. Using such priors for the factor loadings, we obtain the same rate as obtained with point mass mixture priors. To obtain the convergence rates, we construct test functions to separate points in the space of high-dimensional covariance matrices using insights from random matrix theory; the tools developed may be of independent interest.
منابع مشابه
Bayesian Learning in Sparse Graphical Factor Models via Annealed Entropy
We describe a class of sparse latent factor models, called graphical factor models (GFMs), and relevant sparse learning algorithms for posterior mode estimation. Linear, Gaussian GFMs have sparse, orthogonal factor loadings matrices, that, in addition to sparsity of the implied covariance matrices, also induce conditional independence structures via zeros in the implied precision matrices. We d...
متن کاملBayesian Learning in Sparse Graphical Factor Models via Variational Mean-Field Annealing
We describe a class of sparse latent factor models, called graphical factor models (GFMs), and relevant sparse learning algorithms for posterior mode estimation. Linear, Gaussian GFMs have sparse, orthogonal factor loadings matrices, that, in addition to sparsity of the implied covariance matrices, also induce conditional independence structures via zeros in the implied precision matrices. We d...
متن کاملCovariance Kernels from Bayesian Generative Models
We propose the framework of mutual information kernels for learning covariance kernels, as used in Support Vector machines and Gaussian process classifiers, from unlabeled task data using Bayesian techniques. We describe an implementation of this framework which uses variational Bayesian mixtures of factor analyzers in order to attack classification problems in high-dimensional spaces where lab...
متن کاملRate-optimal Posterior Contraction for Sparse Pca By
Principal component analysis (PCA) is possibly one of the most widely used statistical tools to recover a low-rank structure of the data. In the highdimensional settings, the leading eigenvector of the sample covariance can be nearly orthogonal to the true eigenvector. A sparse structure is then commonly assumed along with a low rank structure. Recently, minimax estimation rates of sparse PCA w...
متن کاملar X iv : 1 10 1 . 20 17 v 2 [ st at . M E ] 8 F eb 2 01 1 Bayesian Nonparametric Covariance Regression
Although there is a rich literature on methods for allowing the variance in a univariate regression model to vary with predictors, time and other factors, relatively little has been done in the multivariate case. Our focus is on developing a class of nonparametric covariance regression models, which allow an unknown p × p covariance matrix to change flexibly with predictors. The proposed modeli...
متن کامل